Learning Relation Prototype from Unlabeled Texts for Long-tail Relation Extraction

نویسندگان

چکیده

Relation Extraction (RE) is a vital step to complete Knowledge Graph (KG) by extracting entity relations from texts. However, it usually suffers the long-tail issue. This paper proposes novel approach learn relation prototypes unlabeled texts, facilitate RE transferring knowledge types with sufficient training data. We as an implicit factor between entities, which reflects meanings of and their proximities. construct co-occurrence graph capture both first-order second-order proximities for embedding learning. By optimize distance pairs corresponding prototypes, our method can be easily adapted almost arbitrary frameworks. Thus, learning infrequent or even unseen will benefit semantically proximate through entities large-scale textual information. Extensive experiments on two publicly available datasets present promising improvements (4.1% F1 average). Ablation studies relations, main components, different models demonstrate effectiveness learned prototypes. Finally, we analyze several example cases give intuitive impressions qualitative analysis. Our codes data found in https://github.com/CrisJk/PA-TRP.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Automatic semantic relation extraction from Portuguese texts

Nowadays we are facing a growing demand for semantic knowledge in computational applications, particularly in Natural Language Processing (NLP). However, there aren’t sufficient human resources to produce that knowledge at the same rate of its demand. Considering the Portuguese language, which has few resources in the semantic area, the situation is even more alarming. Aiming to solve that prob...

متن کامل

An Integrated Approach for Relation Extraction from Wikipedia Texts

Linguistic-based methods and web mining-based methods are two types of leading methods for semantic relation extraction task. By integrating linguistic analysis with frequent Web information, this paper presents an unsupervised relation extraction approach, for discovering and enhancing relations in which a specified concept participates. We focus on concepts described in Wikipedia articles. By...

متن کامل

Exploring Long Tail Data in Distantly Supervised Relation Extraction

Distant supervision is an efficient approach for various tasks, such as relation extraction. Most of the recent literature on distantly supervised relation extraction generates labeled data by heuristically aligning knowledge bases with text corpora and then trains supervised relation classification models based on statistical learning. However, extracting long tail relations from the automatic...

متن کامل

CATENA: CAusal and TEmporal relation extraction from NAtural language texts

We present CATENA, a sieve-based system to perform temporal and causal relation extraction and classification from English texts, exploiting the interaction between the temporal and the causal model. We evaluate the performance of each sieve, showing that the rule-based, the machinelearned and the reasoning components all contribute to achieving state-of-the-art performance on TempEval-3 and Ti...

متن کامل

Event and Temporal Relation Extraction from Croatian Newspaper Texts

Event extraction and temporal relation extraction are the subjects of extensive research, which has been additionally motivated by focused evaluation exercises such as TempEval. In this paper we present the work on supervised event and temporal relation extraction from Croatian newspaper texts. Taking into account the limited availability of linguistic tools for Croatian, we focus our research ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Knowledge and Data Engineering

سال: 2021

ISSN: ['1558-2191', '1041-4347', '2326-3865']

DOI: https://doi.org/10.1109/tkde.2021.3096200